Fast Implementation of l 1 Regularized Learning Algorithms Using Gradient Descent Methods ∗
نویسندگان
چکیده
With the advent of high-throughput technologies, l1 regularized learning algorithms have attracted much attention recently. Dozens of algorithms have been proposed for fast implementation, using various advanced optimization techniques. In this paper, we demonstrate that l1 regularized learning problems can be easily solved by using gradient-descent techniques. The basic idea is to transform a convex optimization problem with a non-differentiable objective function into an unconstrained non-convex problem, upon which, via gradient descent, reaching a globally optimum solution is guaranteed. We present detailed implementation of the algorithm using l1 regularized logistic regression as a particular application. We conduct large-scale experiments to compare the new approach with other stateof-the-art algorithms on eight medium and large-scale problems. We demonstrate that our algorithm, though simple, performs similarly or even better than other advanced algorithms in terms of computational efficiency and memory usage. Keyword: l1 regularized learning, feature selection, sparse solution, gradient descent
منابع مشابه
Fast Implementation of ℓ1Regularized Learning Algorithms Using Gradient Descent Methods
With the advent of high-throughput technologies, l1 regularized learning algorithms have attracted much attention recently. Dozens of algorithms have been proposed for fast implementation, using various advanced optimization techniques. In this paper, we demonstrate that l1 regularized learning problems can be easily solved by using gradient-descent techniques. The basic idea is to transform a ...
متن کاملUsing GPUs to speedup sparse coding algorithms applied to self-taught learning problems
In this work, we present a new combination of sparse coding algorithms designed to exploit the GPU and apply it to a self-taught learning [1] problem. We build on top of the iterative approach of solving two convex optimization problems alternatingly (originally suggested in [2]). The first and most important improvement in this work is a parallelized coordinate descent algorithm for the genera...
متن کاملAccelerated parallel and distributed algorithm using limited internal memory for nonnegative matrix factorization
Nonnegative matrix factorization (NMF) is a powerful technique for dimension reduction, extracting latent factors and learning part-based representation. For large datasets, NMF performance depends on some major issues such as fast algorithms, fully parallel distributed feasibility and limited internal memory. This research designs a fast fully parallel and distributed algorithm using limited i...
متن کاملStochastic Coordinate Descent Methods for Regularized Smooth and Nonsmooth Losses
Stochastic Coordinate Descent (SCD) methods are among the first optimization schemes suggested for efficiently solving large scale problems. However, until now, there exists a gap between the convergence rate analysis and practical SCD algorithms for general smooth losses and there is no primal SCD algorithm for nonsmooth losses. In this paper, we discuss these issues using the recently develop...
متن کاملFast Asynchronous Parallel Stochastic Gradient Descent: A Lock-Free Approach with Convergence Guarantee
Stochastic gradient descent (SGD) and its variants have become more and more popular in machine learning due to their efficiency and effectiveness. To handle large-scale problems, researchers have recently proposed several parallel SGD methods for multicore systems. However, existing parallel SGD methods cannot achieve satisfactory performance in real applications. In this paper, we propose a f...
متن کامل